Learning serial constraint-based grammars

نویسنده

  • Robert Staubs
چکیده

In this paper we describe a method for learning grammars in the general framework of Harmonic Serialism (see McCarthy this volume for references and an introduction). We have two main goals. The first is to address the hidden structure problem that serial derivations introduce. The second is to address the problem of learning variation, which has yet to be confronted in this framework (see also Staubs et al. 2010 and Tessier and Jesney 2014 on the learning of Harmonic Serialism). In the remainder of this section, we illustrate the grammar model, and our approach to the learning of hidden structure, with a simple example of stressepenthesis interaction. Our main contribution comes in the next section, where we introduce a method for calculating probabilities over the unbounded non-monotonic derivations that are characteristic of a stochastic serial theory. The third and last section consists of an application to data from French ‘schwa’ deletion that display variation. In Harmonic Serialism, the path from the grammar’s initial input to its final output is a series of derivational steps. In each step, a set of operations first applies to create a candidate set of outputs, from which one is chosen by a set of constraints. This output becomes the input for the next step of the derivation. The derivation terminates, or converges, when the chosen output is identical to the input of that step. From the perspective of learning, these derivations are an instance of what Tesar and Smolensky (2000) call ‘hidden structure’. Hidden structure refers to properties of the learning data that are not supplied to the learner, but must instead be inferred as part of the learning process. Our approach to learning this case of hidden structure is a generalization of Eisenstat’s (2009) method for learning phonological Underlying Representations (URs). We will explain and illustrate Eisenstat’s proposal as we discuss our toy stress-epenthesis case, since it also includes a UR learning problem.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Search Space Properties for Learning a Class of Constraint-based Grammars

We discuss a class of constraint-based grammars, Lexicalized Well-Founded Grammars (LWFGs) and present the theoretical underpinnings for learning these grammars from a representative set of positive examples. Given several assumptions, we define the search space as a complete grammar lattice. In order to prove a learnability theorem, we give a general algorithm through which the top and bottom ...

متن کامل

ML-Tuned Constraint Grammars

In this paper we present a new method for machine learning-based optimization of linguist-written Constraint Grammars. The effect of rule ordering/sorting, grammarsectioning and systematic rule changes is discussed and quantitatively evaluated. The F-score improvement was 0.41 percentage points for a mature (Danish) tagging grammar, and 1.36 percentage points for a half-size grammar, translatin...

متن کامل

Learning OT Constraint Rankings Using a Maximum Entropy Model

A weakness of standard Optimality Theory is its inability to account for grammars with free variation. We describe here the Maximum Entropy model, a general statistical model, and show how it can be applied in a constraint-based linguistic framework to model and learn grammars with free variation, as well as categorical grammars. We report the results of using the MaxEnt model for learning two ...

متن کامل

Richness of the Base and Probabilistic Unsupervised Learning in Optimality Theory

This paper proposes an unsupervised learning algorithm for Optimality Theoretic grammars, which learns a complete constraint ranking and a lexicon given only unstructured surface forms and morphological relations. The learning algorithm, which is based on the ExpectationMaximization algorithm, gradually maximizes the likelihood of the observed forms by adjusting the parameters of a probabilisti...

متن کامل

FCGlight: Making the bridge between Fluid Construction Grammars and main-stream unification grammars by using feature constraint logics

Fluid Construction Grammars (FCGs) are a flavour of Construction Grammars, which themselves are unification-based grammars. Its syntax is (only) up to certain extent similar to other unification-based grammars. However, it lacks a declarative semantics, while its procedural semantics is truly particular, compared to other unification-based grammar formalisms. Here we propose the re-definition o...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2014